141 research outputs found
Wormhole Hamiltonian Monte Carlo
In machine learning and statistics, probabilistic inference involving
multimodal distributions is quite difficult. This is especially true in high
dimensional problems, where most existing algorithms cannot easily move from
one mode to another. To address this issue, we propose a novel Bayesian
inference approach based on Markov Chain Monte Carlo. Our method can
effectively sample from multimodal distributions, especially when the dimension
is high and the modes are isolated. To this end, it exploits and modifies the
Riemannian geometric properties of the target distribution to create
\emph{wormholes} connecting modes in order to facilitate moving between them.
Further, our proposed method uses the regeneration technique in order to adapt
the algorithm by identifying new modes and updating the network of wormholes
without affecting the stationary distribution. To find new modes, as opposed to
rediscovering those previously identified, we employ a novel mode searching
algorithm that explores a \emph{residual energy} function obtained by
subtracting an approximate Gaussian mixture density (based on previously
discovered modes) from the target density function
Spherical Hamiltonian Monte Carlo for Constrained Target Distributions
We propose a new Markov Chain Monte Carlo (MCMC) method for constrained
target distributions. Our method first maps the -dimensional constrained
domain of parameters to the unit ball . Then, it augments the
resulting parameter space to the -dimensional sphere, . The
boundary of corresponds to the equator of . This
change of domains enables us to implicitly handle the original constraints
because while the sampler moves freely on the sphere, it proposes states that
are within the constraints imposed on the original parameter space. To improve
the computational efficiency of our algorithm, we split the Lagrangian dynamics
into several parts such that a part of the dynamics can be handled analytically
by finding the geodesic flow on the sphere. We apply our method to several
examples including truncated Gaussian, Bayesian Lasso, Bayesian bridge
regression, and a copula model for identifying synchrony among multiple
neurons. Our results show that the proposed method can provide a natural and
efficient framework for handling several types of constraints on target
distributions
Earth System Modeling 2.0: A Blueprint for Models That Learn From Observations and Targeted High-Resolution Simulations
Climate projections continue to be marred by large uncertainties, which
originate in processes that need to be parameterized, such as clouds,
convection, and ecosystems. But rapid progress is now within reach. New
computational tools and methods from data assimilation and machine learning
make it possible to integrate global observations and local high-resolution
simulations in an Earth system model (ESM) that systematically learns from
both. Here we propose a blueprint for such an ESM. We outline how
parameterization schemes can learn from global observations and targeted
high-resolution simulations, for example, of clouds and convection, through
matching low-order statistics between ESMs, observations, and high-resolution
simulations. We illustrate learning algorithms for ESMs with a simple dynamical
system that shares characteristics of the climate system; and we discuss the
opportunities the proposed framework presents and the challenges that remain to
realize it.Comment: 32 pages, 3 figure
An Efficient Bayesian Inference Framework for Coalescent-Based Nonparametric Phylodynamics
Phylodynamics focuses on the problem of reconstructing past population size
dynamics from current genetic samples taken from the population of interest.
This technique has been extensively used in many areas of biology, but is
particularly useful for studying the spread of quickly evolving infectious
diseases agents, e.g.,\ influenza virus. Phylodynamics inference uses a
coalescent model that defines a probability density for the genealogy of
randomly sampled individuals from the population. When we assume that such a
genealogy is known, the coalescent model, equipped with a Gaussian process
prior on population size trajectory, allows for nonparametric Bayesian
estimation of population size dynamics. While this approach is quite powerful,
large data sets collected during infectious disease surveillance challenge the
state-of-the-art of Bayesian phylodynamics and demand computationally more
efficient inference framework. To satisfy this demand, we provide a
computationally efficient Bayesian inference framework based on Hamiltonian
Monte Carlo for coalescent process models. Moreover, we show that by splitting
the Hamiltonian function we can further improve the efficiency of this
approach. Using several simulated and real datasets, we show that our method
provides accurate estimates of population size dynamics and is substantially
faster than alternative methods based on elliptical slice sampler and
Metropolis-adjusted Langevin algorithm
Spatiotemporal Besov Priors for Bayesian Inverse Problems
Fast development in science and technology has driven the need for proper
statistical tools to capture special data features such as abrupt changes or
sharp contrast. Many applications in the data science seek spatiotemporal
reconstruction from a sequence of time-dependent objects with discontinuity or
singularity, e.g. dynamic computerized tomography (CT) images with edges.
Traditional methods based on Gaussian processes (GP) may not provide
satisfactory solutions since they tend to offer over-smooth prior candidates.
Recently, Besov process (BP) defined by wavelet expansions with random
coefficients has been proposed as a more appropriate prior for this type of
Bayesian inverse problems. While BP outperforms GP in imaging analysis to
produce edge-preserving reconstructions, it does not automatically incorporate
temporal correlation inherited in the dynamically changing images. In this
paper, we generalize BP to the spatiotemporal domain (STBP) by replacing the
random coefficients in the series expansion with stochastic time functions
following Q-exponential process which governs the temporal correlation
strength. Mathematical and statistical properties about STBP are carefully
studied. A white-noise representation of STBP is also proposed to facilitate
the point estimation through maximum a posterior (MAP) and the uncertainty
quantification (UQ) by posterior sampling. Two limited-angle CT reconstruction
examples and a highly non-linear inverse problem involving Navier-Stokes
equation are used to demonstrate the advantage of the proposed STBP in
preserving spatial features while accounting for temporal changes compared with
the classic STGP and a time-uncorrelated approach.Comment: 29 pages, 13 figure
A Semiparametric Bayesian Model for Detecting Synchrony Among Multiple Neurons
We propose a scalable semiparametric Bayesian model to capture dependencies
among multiple neurons by detecting their co-firing (possibly with some lag
time) patterns over time. After discretizing time so there is at most one spike
at each interval, the resulting sequence of 1's (spike) and 0's (silence) for
each neuron is modeled using the logistic function of a continuous latent
variable with a Gaussian process prior. For multiple neurons, the corresponding
marginal distributions are coupled to their joint probability distribution
using a parametric copula model. The advantages of our approach are as follows:
the nonparametric component (i.e., the Gaussian process model) provides a
flexible framework for modeling the underlying firing rates; the parametric
component (i.e., the copula model) allows us to make inference regarding both
contemporaneous and lagged relationships among neurons; using the copula model,
we construct multivariate probabilistic models by separating the modeling of
univariate marginal distributions from the modeling of dependence structure
among variables; our method is easy to implement using a computationally
efficient sampling algorithm that can be easily extended to high dimensional
problems. Using simulated data, we show that our approach could correctly
capture temporal dependencies in firing rates and identify synchronous neurons.
We also apply our model to spike train data obtained from prefrontal cortical
areas in rat's brain
- …